feat: v2 documentation with versioned navigation and updated SDKs#39
feat: v2 documentation with versioned navigation and updated SDKs#39VinciGit00 wants to merge 36 commits intomainfrom
Conversation
Migrate documentation to v2 structure with versioned nav (v2 default, v1 legacy). Update Python and JavaScript SDK docs to reflect v2 API changes (extract, search, scrape, crawl, monitor namespaces, FetchConfig/LlmConfig). Add v1/ legacy pages with deprecation banners. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
…nd navigation - Add v2 service pages: Extract, Search, Crawl, Monitor - Update Scrape service for v2 format-based API - Update LangChain integration for v2 tools (ExtractTool, SearchTool, etc.) - Update v2 navigation: remove old services (SmartScraper, SearchScraper, Markdownify, SmartCrawler, Sitemap, AgenticScraper, Langflow) Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
- Rename smart-scraper → extract, search-scraper → search - Remove commands dropped from v2: agentic-scraper, generate-schema, sitemap, validate - Update scrape with --format flag (markdown, html, screenshot, branding) - Update crawl with v2 polling model (max-pages, max-depth, max-links-per-page, allow-external) - Update history with v2 service names (scrape, extract, search, monitor, crawl) - Update all examples, JSON mode docs, and AI agent skill docs Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
- Reorder services: Scrape, Extract, Search, Crawl, Monitor - Remove Community anchor, Playground, x402, Langflow from v2 nav - Update Vercel AI integration for v2 SDK Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Update all v2 documentation to match the latest SDK changes: - JS SDK: named import, correct fetchConfig fields (render, wait), maxRetries=2, maxDepth for crawl - Python SDK: Client.from_env(), context manager, format param, history filter params - All service/knowledge-base/cookbook pages: migrate JS examples from v1 individual imports to v2 factory pattern Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Update v2 documentation links to use scrapegrphai.com/dashboard instead of dashboard subdomain and remove the Usage Analytics section from the dashboard overview to match the new content direction. Made-with: Cursor
Refine the docs IA by removing the integrations drawer section and introducing a dedicated transition guide page for users migrating from v1 to v2. Made-with: Cursor
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Rewrite proxy configuration page to document FetchConfig object with mode parameter (auto/fast/js/direct+stealth/js+stealth), country-based geotargeting, and all fetch options. Update knowledge-base proxy guide and fix FetchConfig examples in both Python and JavaScript SDK pages to match the actual v2 API surface. Refs: ScrapeGraphAI/scrapegraph-js#11, ScrapeGraphAI/scrapegraph-py#82 Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
- Rename smart-scraper → extract, search-scraper → search - Remove dropped commands: agentic-scraper, generate-schema, sitemap, validate - Replace --stealth boolean with --mode fetch mode enum - Update scrape with --format flag (markdown, html, screenshot, branding) - Update crawl with v2 polling model and new options - Update env variables to SGAI_API_URL, SGAI_TIMEOUT_S - Update response field names (remainingCredits, markdown) Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Remove the advanced extract accordion snippet to keep the migration-focused section shorter and less confusing. Made-with: Cursor
Clean up the Python SDK docs by removing the non-rendering header image and outdated llm_config parameter rows in extract/search tables. Made-with: Cursor
Add a dedicated Plans & Pricing page and update account docs with current credit costs, proxy modifiers, and plan limits, including navigation links across the section. Made-with: Cursor
Replace legacy v1 code snippets in all use-case pages with actual v2 method names and parameters so examples match the current SDK behavior. Made-with: Cursor
Remove the "What changed at a glance" section from the transition guide as requested, while keeping the detailed migration mapping and examples. Made-with: Cursor
Reorder the v1-to-v2 migration mapping and REST endpoint examples to match the Services navigation sequence (scrape, extract, search, crawl, monitor). Made-with: Cursor
Add concrete Python and JavaScript v2 code examples for the markdownify-to-scrape migration step. Made-with: Cursor
Delete the invalid top image reference on the Mocking & Testing SDK page to avoid broken rendering. Made-with: Cursor
Align monitor docs with Python SDK PR #82 (scrapegraph-py v2): - Rename `cron` parameter to `interval` across all code examples - Add required `name` parameter to SDK snippets - Replace Pydantic model schema with JSON Schema dict - Update FetchConfig usage (mode enum instead of stealth bool) Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Made-with: Cursor
Replace single docs-banner with separate light/dark assets and switch images via Tailwind dark: classes on introduction pages. Made-with: Cursor
Align sdks/javascript.mdx and sdks/python.mdx with the current schemas from scrapegraph-js#11 and scrapegraph-py#82: - search(): add locationGeoCode/location_geo_code, timeRange/time_range, prompt, format, mode; correct numResults default to 3 - extract(): drop llmConfig from params (ignored by v2 route); document mode, contentType, html, markdown alternatives to url - scrape(): document the formats[] array (tagged format entries with per-entry config) and add a multi-format example - crawl.start(): document maxDepth/max_depth, maxPages/max_pages, maxLinksPerPage, allowExternal, contentTypes - monitor.create(): drop prompt (not in v2 schema); add formats and webhookUrl/webhook_url - LlmConfig: clarify it belongs inside scrape json/summary format entries, not on extract/search Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
CLI (just-scrape#13): - scrape: document 8 formats, multi-format via comma-separated -f, and new --html-mode / --scrolls / --prompt / --schema flags - search: document --location-geo-code, --time-range, --format - crawl: document -f / --format - Add the Fetch Modes enum table (auto|fast|js|direct+stealth|js+stealth) that replaces the legacy --stealth boolean MCP server (scrapegraph-mcp#16): - Replace the stale v1 tool list with the v2 surface: markdownify, smartscraper, searchscraper, scrape (formats[]), smartcrawler_* (markdown default), crawl_stop/resume, monitor_* lifecycle, credits, sgai_history - Note removal of sitemap and agentic_scrapper - Document SCRAPEGRAPH_API_BASE_URL override and v2 auth headers Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Group the client-side surfaces (CLI, MCP server, Toonify) with the Python/JavaScript SDKs so they live together in the nav rather than under Services. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
|
Rechecked only the v2 side of this PR against the current
|
Align CLI docs with just-scrape PR #13 (v2 migration): - search: add --location-geo-code, --time-range, --format, --nationality flags - scrape: add links, images, summary, json formats, multi-format, --html-mode, --scrolls, --prompt/--schema - crawl: add --format flag - credits: fix jq path to camelCase remainingCredits - ai-agent-skill: update CLAUDE.md snippet with new formats and flags Align MCP docs with scrapegraph-mcp PR #16 (v2 migration): - Update available tools from 8 v1 tools to 16 v2 tools - Add crawl_stop, crawl_resume, credits, sgai_history, monitor_* tools - Remove sitemap, agentic_scrapper references Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Rewrite all JavaScript code examples to match the new v2 SDK API from ScrapeGraphAI/scrapegraph-js#13. Key changes: - Replace factory pattern (scrapegraphai({ apiKey })) with direct imports - All functions use (apiKey, params) signature - scrape() uses formats array instead of single format string - Return type is ApiResult<T> with status check, not throw-on-error - crawl.status() renamed to crawl.get(), crawl.delete() added - monitor.create() uses formats array, not prompt - Restore generateSchema and checkHealth in docs - Schema params use JSON objects, not Zod instances - history is now history.list() and history.get() Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
- Add Monitor command section (create, list, get, update, delete, pause, resume) - Fix extract: --mode is HTML mode (normal/reader/prune), not fetch mode - Fix search: --location-geo-code → --country - Fix fetch modes: auto|fast|js + separate --stealth flag - Add schema to history services list Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
scrapegraph-py v2 is a complete rewrite (Pydantic v2 + httpx + ApiResult wrapper + nested resources). Align the docs with the new SDK surface: Python SDK (sdks/python.mdx): - Rename Client -> ScrapeGraphAI, AsyncClient -> AsyncScrapeGraphAI - Show typed Request models (ScrapeRequest, ExtractRequest, SearchRequest, CrawlRequest, MonitorCreateRequest, HistoryFilter) - Document ApiResult[T] pattern (status / data / error / elapsed_ms) - Document Pydantic format classes (MarkdownFormatConfig, JsonFormatConfig, ScreenshotFormatConfig, etc.) - Document nested resources: sgai.crawl.*, sgai.monitor.*, sgai.history.* - Document SGAI_API_KEY / SGAI_API_URL / SGAI_TIMEOUT_S / SGAI_DEBUG env vars - Drop removed methods (markdownify, agenticscraper, sitemap, feedback) MCP (services/mcp-server.mdx, services/mcp-server/introduction.mdx): - Base URL: /api/v2 -> /v2 (matches scrapegraph-mcp PR #16 realignment) - Env vars: SGAI_API_URL / SGAI_TIMEOUT_S (SCRAPEGRAPH_API_BASE_URL legacy) - Auth: SGAI-APIKEY only (dropped Authorization: Bearer to match SDK) - Document generate_schema tool - Expand smartcrawler_initiate formats (markdown/html/links/images/summary/ branding/screenshot) - Document FetchConfig passthrough params on all content tools - Monitor: clarify prompt + output_schema wrap into json format entry Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Brings the CLI docs in line with the CLI changes in ScrapeGraphAI/just-scrape#14 (which pulls in scrapegraph-js v2 PR #13 head 096c110): - Document the full `just-scrape monitor` action set, including the new `monitor activity --id <id> [--limit] [--cursor]` for paginated tick history - Replace stale `-m direct+stealth` / `-m js+stealth` with real CLI syntax (`-m js --stealth`, fetch modes: auto/fast/js) - Env vars: `SGAI_TIMEOUT_S` → `SGAI_TIMEOUT`, default base URL now `https://api.scrapegraphai.com/api/v2`, document `SGAI_DEBUG` - Credits example uses `.remaining` (v2 response shape) - Add `schema` to the history services list - Fix `--location-geo-code` → `--country` in search example - Add monitor usage examples (webhook, activity, jq filter for changes) Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Summary
docs.jsonnow usesnavigation.versionswith v2 as default and v1 as legacy, matching the structure from doc-privateextract,search,scrape,crawl.*,monitor.*,credits,history),FetchConfig/LlmConfigconfig objects, factory pattern for JSunittest.mock,responses, Jest/Vitest, MSW)v1/with deprecation banners pointing to v2 equivalentsTest plan
🤖 Generated with Claude Code